Dynamic Affine-Invariant Shape-Appearance Handshape Features and Classification in Sign Language Videos

نویسندگان

  • Anastasios Roussos
  • Stavros Theodorakis
  • Vassilis Pitsikalis
  • Petros Maragos
چکیده

We propose the novel approach of dynamic affine-invariant shape-appearance model (Aff-SAM) and employ it for handshape classification and sign recognition in sign language (SL) videos. AffSAM offers a compact and descriptive representation of hand configurations as well as regularized model-fitting, assisting hand tracking and extracting handshape features. We construct SA images representing the hand’s shape and appearance without landmark points. We model the variation of the images by linear combinations of eigenimages followed by affine transformations, accounting for 3D hand pose changes and improving model’s compactness. We also incorporate static and dynamic handshape priors, offering robustness in occlusions, which occur often in signing. The approach includes an affine signer adaptation component at the visual level, without requiring training from scratch a new singer-specific model. We rather employ a short development data set to adapt the models for a new signer. Experiments on the Boston-University-400 continuous SL corpus demonstrate improvements on handshape classification when compared to other feature extraction approaches. Supplementary evaluations of sign recognition experiments, are conducted on a multi-signer, 100-sign data set, from the Greek sign language lemmas corpus. These explore the fusion with movement cues as well as signer adaptation of Aff-SAM to multiple signers providing promising results.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Hand Tracking and Affine Shape-Appearance Handshape Sub-units in Continuous Sign Language Recognition

We propose and investigate a framework that utilizes novel aspects concerning probabilistic and morphological visual processing for the segmentation, tracking and handshape modeling of the hands, which is used as front-end for sign language video analysis. Our ultimate goal is to explore the automatic Handshape Sub-Unit (HSU) construction and moreover the exploitation of the overall system in a...

متن کامل

Advances in Dynamic-Static Integration of Movement and Handshape Cues for Sign Language Recognition

We explore the integration of movement-position (MP) and handshape (HS) cues for sign language recognition. The proposed method combines the data-driven subunit (SU) modeling exploiting the dynamicstatic notion for MP and the affine shape-appearance SUs for HS configurations. These aspects lead to the new dynamic-static integration of manual cues. This data-driven scheme takes advantage of the ...

متن کامل

Extraction of Isolated Signs from Sign Language Videos via Multiple Sequence Alignment

In this work, we present an alignment-based method to perform sign segmentation and to extract isolated signs from continuous sign language videos. Sign videos contain many modalities, the most prominent of which are hand gestures, manifested as hand motion and shape, which are represented by a variety of extracted features in this work. We compare two different alignment approaches, Dynamic Ti...

متن کامل

The Importance of 3D Motion Trajectories for Computer-based Sign Recognition

Computer-based sign language recognition from video is a challenging problem because of the spatiotemporal complexities inherent in sign production and the variations within and across signers. However, linguistic information can help constrain sign recognition to make it a more feasible classification problem. We have previously explored recognition of linguistically significant 3D hand config...

متن کامل

Experiments on Global and Local Active Appearance Models for Analysis of Sign Language Facial Expressions

We explore features based on Active Appearance Modeling (AAM) of facial images within sign language videos. We employ a global AAM that initializes multiple local AAMs around places of interest. The local features offer a compact and descriptive representation of the facial regions of interest. The Global and Local AAM (GLAAM) is applied on Sign Language videos, and evaluated on classification ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2013